Published on in Vol 5 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/80348, first published .
Large Language Model–Based Chatbots and Agentic AI for Mental Health Counseling: Systematic Review of Methodologies, Evaluation Frameworks, and Ethical Safeguards

Large Language Model–Based Chatbots and Agentic AI for Mental Health Counseling: Systematic Review of Methodologies, Evaluation Frameworks, and Ethical Safeguards

Large Language Model–Based Chatbots and Agentic AI for Mental Health Counseling: Systematic Review of Methodologies, Evaluation Frameworks, and Ethical Safeguards

Authors of this article:

Ha Na Cho1 Author Orcid Image ;   Jiayuan Wang1 Author Orcid Image ;   Di Hu1 Author Orcid Image ;   Kai Zheng1 Author Orcid Image

Review

Department of Informatics, University of California, Irvine, Irvine, CA, United States

Corresponding Author:

Ha Na Cho, MS

Department of Informatics

University of California, Irvine

5029 Donald Bren Hall

Irvine, CA, 92697

United States

Phone: 1 949 400 0992

Email: chohn1@uci.edu


Background: Large language model (LLM)–based chatbots have rapidly emerged as tools for digital mental health (MH) counseling. However, evidence on their methodological quality, evaluation rigor, and ethical safeguards remains fragmented, limiting interpretation of clinical readiness and deployment safety.

Objective: This systematic review aimed to synthesize the methodologies, evaluation practices, and ethical or governance frameworks of LLM-based chatbots developed for MH counseling and to identify gaps affecting validity, reproducibility, and translation.

Methods: We searched Google Scholar, PubMed, IEEE Xplore, and ACM Digital Library for studies published between January 2020 and May 2025. Eligible studies reported original development or empirical evaluation of LLM-driven MH counseling chatbots. We excluded studies that did not involve LLM-based conversational agents, were not focused on counseling or supportive MH communication, or lacked evaluable system outputs or outcomes. Screening and data extraction were conducted in Covidence (Veritas Health Innovation) following PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) 2020 guidelines. Study quality was appraised using a structured traffic light framework across 5 methodological domains (design, dataset reporting, evaluation metrics, external validation, and ethics), with an overall judgment derived across domains. We used narrative synthesis with descriptive aggregation to summarize methodological trends, evaluation metrics, and governance considerations.

Results: Twenty studies met the inclusion criteria. GPT-based models (GPT-2/3/4) were used in 45% (9/20) of studies, while 90% (18/20) used fine-tuned or domain-adaptation models such as LLaMa, ChatGLM, or Qwen. Reported deployment types were not mutually exclusive; standalone apps were most common (18/20, 90%), and some systems were also implemented as virtual agents (4/20, 20%) or delivered via existing platforms (2/20, 10%). Evaluation approaches were frequently mixed, with qualitative assessment (13/20, 65%), such as thematic analysis or rubric-based scoring, often complemented by quantitative language metrics (18/20, 90%), including BLEU (Bilingual Evaluation Understudy), ROUGE (Recall-Oriented Understudy for Gisting Evaluation), or perplexity. Quality appraisal indicated consistently low risk for dataset reporting and evaluation metrics, but recurring limitations were observed in external validation and reporting on ethics and safety, including incomplete documentation of safety safeguards and governance practices. No included study reported registered randomized controlled trials or independent clinical validation in real-world care settings.

Conclusions: LLM-based MH counseling chatbots show promise for scalable and personalized support, but current evidence is limited by heterogeneous study designs, minimal external validation, and inconsistent reporting of safety and governance practices. Future work should prioritize clinically grounded evaluation frameworks, transparent reporting of model and prompt configurations, and stronger validation using standardized outcomes to support safe, reliable, and regulatory-ready deployment.

JMIR AI 2026;5:e80348

doi:10.2196/80348

Keywords



The increasing global burden of mental health (MH) disorders necessitates accessible and scalable intervention strategies [1]. Persistent challenges, such as provider shortages, access disparities, and delayed care, continue to limit timely support for individuals in need. In this context, digital mental health interventions (DMHIs) have emerged as potential tools to complement existing services, particularly for supportive counseling and early intervention [2]. Among these, large language model (LLM)–based chatbots have emerged as effective tools for delivering scalable psychological counseling and intervention strategies. These artificial intelligence (AI)–driven systems can respond instantly, adapt to user needs, and offer personalized support, making them a valuable resource for individuals who lack immediate access to human experts. However, despite the growing interest in the DMHI domain, concerns and challenges remain regarding their effectiveness and reliability in clinical practice. Addressing these limitations requires a comprehensive assessment of current LLM-based chatbot methodologies, their evaluation frameworks, and their role in MH counseling.

Previous systematic reviews have explored the effectiveness of rule-based and traditional AI-driven chatbots for DMHI. Early studies primarily focused on text-based conversational agents that followed prescripted dialogues or used traditional machine learning models trained on structured datasets [3,4]. These reviews highlight the potential of AI-based DMHI to enhance user engagement and symptom management, while also exposing key limitations such as rigid responses, limited adaptability, and a lack of contextual understanding [5,6]. More recent reviews have examined LLMs’ integration in MH, demonstrating that LLM-based chatbots exhibit superior natural language understanding, response flexibility, and contextual understanding compared to earlier MH chatbot architectures [7-9]. However, many of these reviews adopt a broad focus, assessing LLM applications in MH for education, screening, and assessment, rather than examining their role in counseling-specific interventions [10,11]. This gap leaves uncertainty regarding the effectiveness and clinical applicability of LLMs explicitly designed for MH counseling. Despite the increasing adoption of LLMs in digital therapy, existing studies have yet to systematically assess their clinical effectiveness, response accuracy, and real-world impact in MH counseling scenarios [12]. Despite their promise, LLM-based counseling systems raise substantial ethical and safety concerns, including biased responses, hallucinated information, inappropriate reassurance, and potential harm when interacting with emotionally vulnerable users. These risks motivate the need for systematic evaluation of both methodological rigor and ethical safeguards.

Recent advances in MH-focused LLMs demonstrated substantial improvements in natural language processing (NLP), contextual awareness, and response generation. Transformer-based models such as GPT series, LLaMa, MedAlpaca [13], and ChatDoctor [14] have been increasingly integrated for MH applications, offering real-time user interaction and context-aware counseling support. These models can analyze user queries, infer emotional states, and generate human-like responses, providing automated counseling services to users. Moreover, fine-tuned LLMs tailored to MH applications, notably MentalLlama [15], Mistral [16], and MentalBART [17], have shown promise in outperforming general models on key counseling metrics. Despite these advancements, existing models still face significant challenges, including biases in training data, potential misinformation, lack of clinical validation, and ethical concerns related to overreliance on AI-generated counseling [18]. Furthermore, rigorous evaluation frameworks to ensure clinical efficacy, safety, and real-world applicability remain limited, highlighting the need for standardized assessment methodologies in MH-focused LLM deployment [19].

While recent reviews have LLMs in MH broadly, most adopt a wide scope encompassing screening, education, and assessment tasks. In contrast, this review focuses specifically on LLM-based systems designed for counseling and supportive dialogue, where conversational appropriateness, safety, and ethical governance are central concerns. We systematically examine how these counseling-oriented systems are developed and evaluated, including model preprocessing, prompt engineering, fine-tuning strategies, deployment formats, and reported safeguards against harm. By synthesizing methodological, evaluative, and ethical patterns across studies, this review provides a focused analysis of the current landscape of LLM-driven MH counseling and identifies key gaps relevant to future research and clinical integration. Accordingly, this review was guided by three primary questions:

  1. How have LLMs been applied and evaluated for MH counseling or support tasks?
  2. What types of clinical or psychometric outcomes have been measured to assess their effectiveness and safety?
  3. How have studies addressed ethical considerations such as bias, misinformation, and user safety when deploying LLM-based chatbots or agentic systems?

Search Strategy

This study followed PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines, ensuring a structured and transparent review process. We conducted a systematic literature search across Google Scholar, PubMed, IEEE Xplore, and ACM Digital Library, covering studies published between January 2020 and May 2025. The search queries incorporated LLM-related terms (“large language model” OR “LLM” OR “LLMs”) and MH-specific terms (“mental illness” OR “mental disorder” OR “mental health” OR “mental wellness”), along with intervention-focused keywords (“chatbot” OR “counseling” OR “conversational” OR “digital intervention” or “DMHI”). The final search queries were iteratively refined using Boolean logic and expert input and were performed on May 15, 2025, and study screening and data extraction were completed on June 10, 2025. The full database-specific search queries are provided in Multimedia Appendix 1 and operational definitions of key technical terms used for coding and synthesis are provided in Multimedia Appendix 2. The PRISMA checklist is provided in Multimedia Appendix 3.

Study screening, data extraction, and methodological evaluation were conducted collaboratively by 3 reviewers (HC, JW, and DH) using Covidence (Veritas Health Innovation) and Zotero (Corporation for Digital Scholarship, CDS). The reviewers independently screened titles, abstracts, and full texts to determine study eligibility based on predefined inclusion and exclusion criteria. Discrepancies in study selection and data extraction were discussed during weekly meetings, where consensus was reached through deliberation. Any disagreements that persisted were resolved through majority voting, ensuring consistency and rigor in the systematic review process.

The initial search retrieved 1593 studies, which underwent a multistage screening process, including duplication removal, title and abstract review screening, followed by full-text analysis. After applying inclusion and exclusion criteria, a total of 20 studies [20-39] were included in the final review. Figure 1 presents the conceptual framework of this systematic review, illustrating how extracted methodological evidence is synthesized and translated into downstream design and evaluation considerations for LLM-based MH systems.

Figure 1. Conceptual framework of the systematic review. LLM: large language model; NLP: natural language processing.

Inclusion and Exclusion Criteria

Studies were included if they evaluated LLM-based chatbots for MH counseling and reported performance measures on response quality, engagement, effectiveness, or ethical considerations, with only English-language studies considered. Only original peer-reviewed studies published between 2020 and 2025 were considered. Studies implementing fine-tuning, prompt engineering, or structured evaluation methodologies were included for comparative analysis. Studies were excluded if a study design misaligned with the research objectives, focused on rule-based chatbots, assessed general-purpose conversational AI without an MH focus, or lacked empirical evaluation. Review articles, opinion pieces, and non–peer-reviewed sources were also excluded. In addition, preprints and conference proceedings were screened when they reported original empirical evaluations of LLM-based MH counseling systems and provided sufficient methodological detail to permit appraisal. Preprints were included selectively to capture emerging methodological approaches in this rapidly evolving field; however, non–peer-reviewed records with insufficient methodological detail were excluded, and preprints were interpreted with appropriate caution alongside peer-reviewed evidence.

Data Collection and Extraction

This review extracted key model attributes, development methodologies, evaluation frameworks, and study outcomes across the studies on LLM-based MH chatbots. Extracted data fields included LLM and version (eg, GPT-4, LLaMA, and ChatGLM2), given model name, and service approach (eg, standalone chatbot, integrated platform, mobile app, or virtual LLM agent). Additionally, we gathered details on dataset characteristics, distinguishing between studies that used publicly available datasets, proprietary datasets, or synthetically generated datasets for fine-tuning. Preprocessing techniques were identified, including text normalization, augmentation, and filtering. For model development, we extracted information on prompting strategies, as well as fine-tuning methods. Evaluation methodologies were categorized into qualitative and quantitative approaches. We identified human evaluation metrics, such as expert-reviewed response accuracy, empathy, and coherence, along with automated quantitative evaluation methods (eg, BLEU [Bilingual Evaluation Understudy] and ROUGE [Recall-Oriented Understudy for Gisting Evaluation]). Furthermore, the review examined whether LLM-generated responses were fully automated, human-in-the-loop, or used for backend knowledge retrieval. Notably, not all studies provided information across all extracted fields, and when data were unavailable, it was marked as “NA.”

Given the technical focus of this review, outcome measures were extracted primarily in relation to model performance and evaluation methodology rather than user-reported or clinical outcomes. Where available, outcomes were categorized as user-reported or expert-assessed for clarity, though most studies emphasized model-level performance metrics.

Risk of Bias and Study Quality

All screening, data extraction, and quality appraisal procedures were conducted in Covidence in accordance with PRISMA 2020 guidelines. Risk of bias was assessed using a structured traffic light framework covering 5 methodological domains: design, dataset reporting, evaluation metrics, external validation, and ethics, with an overall judgment summarized across domains. Detailed visualizations of study-level judgments are presented in Figures 2 and 3.

Figure 2. Summary of risk of bias across domains.

Two reviewers independently performed the assessments and resolved discrepancies (<5%) through discussion until consensus was reached. Domain-level ratings were assigned using study-reported methodological details, including eligibility criteria, dataset characteristics, predictor specification, outcome definitions, and validation strategies. Standardized Covidence extraction forms were used to support consistent evaluation across all included studies.

Across the design domain, most studies, such as Lai et al [35], Marmol-Romero et al [20], Chen et al [37], and George et al [22], were rated as low risk, with Moon et al [31], Gu and Zhu [28], Xiao et al [38], Kang and Hong [23], Qiu et al [24], Hu et al [39], and Gaikwad et al [33] resulting in some concerns (Figure 3) [20-39]. Evaluation metrics were similarly robust across the literature, with almost all studies receiving low-risk ratings, except Moon et al [31] and Gu and Zhu [28] showed some concerns. External validation represented the most prominent methodological limitation, with high risk identified for Yu and McGuinness [29] and Na [36]. Ethics-related reporting showed the greatest concentration of high-risk judgments, with Moon et al [31], Lai et al [35], Marmol-Romero et al [20], Chen et al [37], Agnihotri et al [21], and George et al [22] rated as high risk, while some concerns were observed for Xiao et al [38], Kang and Hong [23], Qiu et al [24], Gu and Zhu [28], Gaikwad et al [33], Hu et al [39], Mavila et al [34], and Hu et al [30]. Overall judgments reflected these patterns, indicating that while core reporting of datasets and evaluation metrics was generally adequate, gaps in external validation and ethics transparency remain key weaknesses in the current evidence base.

Figure 3. Traffic light plot for the assessment of risk of bias [20-39].

Ethical Considerations

Ethical risks, including bias propagation, hallucinated outputs, and misinformation, were largely underexplored. Only 3 studies [28,31,35] briefly mentioned potential harms, and none systematically audited their models for safety in high-risk user scenarios. No study documented mitigation strategies for hallucinations or unintended outputs. Furthermore, training data provenance and annotation processes were rarely disclosed, making it difficult to assess content, privacy, or representational bias. Across the included studies, explicit ethical deployment frameworks, alignment with international AI governance guidelines (eg, World Health Organization [WHO] and Organisation for Economic Co-operation and Development [OECD]), or formal safety auditing protocols were rarely described, particularly in relation to vulnerable user populations.


Study Characteristics

Our systematic review identified 20 studies [20-39] focusing on LLM-based MH chatbots and digital interventions (Figure 4). Notably, 95% (19/20) of the studies were published between 2023 and 2025, reflecting a rapid acceleration of research in this area. The distribution of publication venues varied, with 8 studies [1,4,7,9,12,15,18,20] appearing in MH-focused journals, 9 studies [2,3,5,6,8,10,11,14,17] in computer science conferences and journals, and 3 studies [13,16,19] in interdisciplinary domains such as medical informatics and AI applications in DMHI (Table 1). While some studies spanned multiple domains, each was counted once based on its primary publication type.

Figure 4. PRISMA 2020 flow diagram summarizing the identification, screening, and inclusion process for studies on large language model–based mental health counseling systems.
Table 1. Overview of LLMa-based mental health chatbots included in the review, summarizing study year, model name, underlying architecture, prompting strategy, and fine-tuning approach.
Author (year)Model nameIntegrated LLMPrompt methodFine-tuning
Moon et al [31] (2024)bMistral-7B Instruct-v0.2Few-shot, CoTc, RAGd, and instruction tuningTask-specific FTe and FT-PEFTf (QLoRAg)
Lai et al [35] (2023)Psy-LLMPanGu, Wen-ZhongFew-shot, RAG, instruction tuning, and persona-basedTask-specific FT, FT-PEFT (LoRAh), instruction FT, and continual learning
Marmol-Romero [20] et al (2024)GPT-3Few-shot, prompt chaining, instruction tuning, and persona-based
Chen et al [37] (2024)PBChatChatGLM2Few-shot, CoT, prompt chaining, instruction tuning, and persona-basedTask-specific FT and FT-PEFT (QLoRA)
Agnihotri et al [21] (2021)TACAGPT-2Few-shot, CoT, RAG, prompt chaining, instruction tuning, and persona-basedTask-specific FT, full FT, and instruction FT
George et al [22] (2024)MelloMistral-7B Instruct v0.1Few-shot, prompt chaining, instruction tuning, and persona-basedTask-specific FT, FT-PEFT (QLoRA), and instruction FT
Xiao et al [38] (2024)HealMeLLaMA2-7b-chatFew-shot, CoT, prompt chaining, instruction tuning, and persona-basedSupervised fine-tuning and LoRA
Kang et al [23] (2025)HoMemeTown Dr. CareSamGPT-4.0Few-shot, RAG, prompt chaining, instruction tuning, and persona-based
Qiu et al [24] (2024)PsyChatChatGLM2-6B2Few-shot, CoT, RAG, prompt chaining, instruction tuning, and persona-basedTask-specific FT, FT-PEFT (LoRA), and instruction FT
Hanji et al [32]
(2024)
Self-HealGPT-2Few-shot, prompt chaining, instruction tuning, persona-based, and multimodal inputTask-specific FT, full FT, and instruction FT
Kang et al [25]
(2024)
LLaMA2-7B, ChatGLM2-6BFew-shot, CoT, RAG, prompt chaining, and instruction tuningTask-specific FT, FT-PEFT (inhibited LoRA), and RLHFi
Fan et al [26]
(2025)
Qwen2-7BFew-shot, CoT, RAG, prompt chaining, instruction tuning, and persona-basedTask-specific FT, FT-PEFT (LoRA), instruction FT (GPT-4 assisted), and RLHF
Mishra et al [27]
(2023)
eTHERAPISTGPT-2 MediumFew-shot, prompt chaining, instruction tuning, and persona-basedTask-specific FT, full FT, instruction FT, and RLHF
Gu and Zhu [28] (2024)MentalBlendGPT-4, GPT-3.5-turboFew-shot, CoT, RAG, prompt chaining, instruction tuning, and persona-basedTask-specific FT
Gaikwad et al [33]
(2024)
Sahara: Virtual CompanionDialoGPT + T5Few-shot, RAG, prompt chaining, instruction tuning, persona-based, and multimodal inputTask-specific FT, full FT, and instruction FT
Na et al [36] (2024)CBT-LLMBaichuan-7BFew-shot, CoT, RAG, prompt chaining, instruction tuning, and persona-basedTask-specific FT, FT-PEFT (LoRA), and instruction FT
Yu and McGuinness et al [29] (2024)DialoGPT, ChatGPT 3.5Few-shot, RAG, prompt chaining, instruction tuning, and persona-basedTask-specific FT and full FT
Hu et al [39] (2024)Qwen-7B, Qwen-maxFew-shot, CoT, RAG, prompt chaining, instruction tuning, persona-based, and multimodal inputTask-specific FT, FT-PEFT (LoRA), and instruction FT
Mavila et al [34] (2024)iCareRASA, GPT-3Few-shot, RAG, prompt chaining, instruction tuning, and persona-basedTask-specific FT and instruction FT
Hu et al [30]
(2024)
PsycoLLMQwen1.5-14B-ChatFew-shot, RAG, prompt chaining, instruction tuning, and persona-basedFull FT, task-specific FT, and instruction FT

aLLM: large language model.

bNot available.

cCoT: chain-of-thought.

dRAG: retrieval-augmented generation.

eFT: fine-tuning.

fPEFT: parameter-efficient fine-tuning.

gQLoRA: quantized low-rank adaptation.

hLoRA: low-rank adaptation.

iRLHF: reinforcement learning with human feedback.

Model Overview

The reviewed studies varied in their application focus and deployment of LLM-based MH systems. Among the 20 studies [20-39], the majority (n=12) developed models aimed at general MH support, including conversational counseling, self-guided emotional regulation, and psychoeducational engagement. These included Marmol-Romero et al [20], TACA [21], Mello [22], HoMemeTown Dr. CareSam [23], PsyChat [24], Kang et al [25], Fan et al [26], e-THERAPIST [27], MentalBlend [28], Yu and McGuinness et al [29], PsycoLLM [30], and PBChat [37].

Six studies designed systems targeting specific psychological conditions such as depression, anxiety, or stress-related symptoms. For instance, Moon et al [31], Self-Heal [32], Sahara [33], iCare [34], Psy-LLM [35], and CBT-LLM [36] were explicitly grounded in condition-specific goals, with some aligned to structured therapeutic frameworks such as cognitive behavioral therapy (CBT). A smaller set of studies addressed niche applications, including behavioral pattern recognition (HealMe [38]) and child mental resilience (Hu et al [39]).

Notably, only 6 [21,22,24,31,32,38] out of 20 studies [20-39] provided open-source code or implementation details. The remaining 14 [20,23,25-30,33-37,39] studies did not release their models, limiting reproducibility and transparency. This pattern was associated with limited opportunities for independent benchmarking, external validation, and real-world adaptation of LLM-based digital MH tools.

Dataset Sources and Preprocessing

The dataset sources and preprocessing methods varied significantly among the studies. A total of 13 studies used publicly available datasets, such as HOPE and MotiVAte Dataset (Moon et al [31]), ScenarioSA (TACA [21]), Counsel-Chat (Mello [22], iCare [34]), PATTERNREFRAME (HealMe [38]), SmileChat and Xingling (PsyChat [24]), Therapy_data (Self-Heal [32]), Alexander Street Therapy Data (Kang et al [23]), Emotional First Aid and Emotional-Support-Conversation Dataset (Fan et al [26]), PsyQA (Psy-LLM [35], CBT-LLM [36], MentalBlend [28]), and the Mental Health Corpus (Sahara: Virtual Companion [33]), enabling greater replicability. Six studies [20,25,27,29,30,39] relied on proprietary or unspecified datasets, including therapy session notes, real-world therapeutic conversations, and psychological counseling transcripts.

Text preprocessing was applied in 16 studies [20-24,26-36], primarily involving normalization, token cleanup, and content filtering to improve model robustness. Six studies [21,22,24,31,32,38] additionally used data augmentation techniques such as paraphrasing and backtranslation to enhance linguistic variability and generalizability. Tokenization methods varied as well: standard subword tokenization (eg, Byte Pair Encoding and WordPiece) was applied in 13 studies [20-22,24-26,28-33,37], while domain-specific tokenizers were implemented in 3 studies [33,35,36], particularly those working with non-English or clinical corpora.

Model Architectures

The LLM architectures used across the 20 studies [20-39] reflected a range of general-purpose base models, instruction-tuned variants, and regionally adapted LLMs, rather than models inherently designed for MH applications. GPT-family models (GPT-2, GPT-3, GPT-3.5, and GPT-4) were the most commonly used, appearing in 9 studies [20-22,24,26,29-31,37], making them the dominant decoder-only generative architecture for counseling-oriented dialogue generation. These models were deployed in either zero-shot or fine-tuned configurations, depending on study objectives.

Beyond GPT, LLaMA-based models were used in 2 studies [28,35], and Mistral-based instruction-tuned models appeared in another 2 [28,35]. Importantly, Mistral itself is a general-purpose LLM and not intrinsically tailored to MH tasks; MH specificity in these studies was introduced through task-specific fine-tuning, prompt design, or alignment strategies rather than through the base architecture. Several studies also adopted region-specific LLMs for multilingual or non-English settings, including ChatGLM (3 studies [31,33,39]), Qwen (3 studies [25,34,38]), and Baichuan (1 study [26]), with applications often focused on Chinese-speaking populations. Additional generative models included DialoGPT and T5, which were selected for their conversational or encoder-decoder generation capabilities, while RASA was used primarily as a rule-based or hybrid dialogue management framework rather than a large generative model.

Importantly, no study used encoder-only architectures such as Bidirectional Encoder Representations from Transformers (BERT) as the core model for dialogue generation. Consistent with their design, encoder-based transformers are typically applied to classification, retrieval, or extractive question answering tasks, rather than open-ended response generation. The reviewed studies instead demonstrate a clear preference for decoder-only or encoder-decoder generative architectures that support few-shot prompting, instruction tuning, and flexible conversational output, which are better aligned with interactive MH counseling applications.

Prompt Engineering

Prompt engineering was frequently reported as part of model design and evaluation across studies, though its independent impact on performance was not systematically quantified (Table 1). Few-shot prompting was used in 18 [20-37] of the 20 studies [20-39], providing models with minimal yet targeted examples to guide response generation. Instruction tuning, which aligns the model’s responses with therapist-like behavior or task-specific expectations, was found in 19 studies [20-33,35-39], reflecting its dominance in aligning LLM behavior with MH intervention goals.

Persona-based prompting was implemented in 18 studies [20-24,26-33,35-39] to simulate consistent, empathetic counselor personas. Prompt chaining, also used in 18 studies [20-29,31-33,35-39], enabled multistep reasoning and follow-up generation logic. Retrieval-augmented generation (RAG) was used in 15 studies [21-24,26-36] to incorporate external knowledge sources, such as clinical guidelines or internal documentation, into the response generation process. Chain-of-thought (CoT) prompting, designed to improve reasoning quality, was applied in 9 studies [22,24,26,29,31,33,35-37]. Only 3 studies [26,31,39] explored multimodal input prompting, indicating its nascent use in this space.

Fine-Tuning Approaches

Fine-tuning strategies varied in scope and efficiency (Table 1). Task-specific fine-tuning was the most common, used in 18 studies [20-37] to tailor model behavior toward specific therapeutic objectives such as empathy generation, goal setting, or motivational interviewing. Instruction fine-tuning was reported in 12 studies [20-24,26,28-31,33,35], reinforcing alignment with counseling frameworks. Parameter-efficient fine-tuning (PEFT), including low-rank adaptation (LoRA) and its quantized variant quantized low-rank adaptation (QLoRA), was applied in 10 studies [21,22,24,26,28,30-32,35,38], often motivated by computational constraints or the need for modular updates.

Full fine-tuning, which requires more extensive retraining of the model, was performed in 6 studies [20,23,27,29,34,37]. Although computationally intensive, it enabled deeper customization. Reinforcement Learning from Human Feedback (RLHF) was implemented in 3 studies [31,33,36] to further refine model behavior based on human preferences and contextual alignment. Continual learning, the ability to incrementally update the model as new data become available, was mentioned in only one study [39], suggesting it remains an underexplored area in LLM-based MH systems.

Evaluation Metrics

Evaluation methods across the included studies spanned both qualitative and quantitative approaches (Table 2). Human evaluation was reported in 18 [20-37] of the 20 studies [20-39], with assessments commonly carried out by MH professionals, trained researchers, or domain experts. These evaluations focused on response quality, contextual relevance, empathetic tone, and overall helpfulness. Five studies went further by evaluating subjective dimensions such as perceived trust, safety, and user comfort, reinforcing the complexity and sensitivity of human-chatbot interactions in MH contexts.

A total of 18 studies [20-31,33-38] incorporated quantitative evaluation methods, reflecting growing efforts to standardize outcome measurement. Among these, lexical overlap metrics were the most frequently reported: BLEU, ROUGE, and distinct-n scores were used in 11 studies [24,25,27-31,33,35-37] to assess textual similarity or diversity between model-generated responses and reference utterances. These metrics, while commonly applied in natural language generation, provide limited insight into conversational appropriateness or therapeutic value. Six studies [24,27,29,33,35,36] used fluency and coherence metrics, including perplexity and linguistic entropy, to quantify the structural quality and readability of generated responses. Despite the emphasis on linguistic metrics, only a small subset used validated psychometric instruments, such as depression screening scales or evidence-based dialogue frameworks. This indicates a significant gap in the clinical grounding of current evaluation pipelines. No studies reported the use of instruments such as the Patient Health Questionnaire-9 (PHQ-9), Generalized Anxiety Disorder-7 (GAD-7), or System Usability Scale (SUS) in a standardized clinical setting, and no study reported results from a registered or randomized controlled trial.

Table 2. Evaluation metrics used across included studies, distinguishing between automated NLPa-based measures and human-assessed qualitative metrics.
Author (year)Evaluation metrics (automatic)Evaluation metrics (human)Human evaluation (N)
Moon et al [31] (2024)F1-score, BLEUb, ROUGE-Lc, and BERTScoredHuman/Expert (activation event, beliefs, consequences, and distorted parts)Annotators (n=2)
Lai et al [35] (2023)Perplexity, ROUGE-L, and distinct-n metricsHuman/Expert (helpfulness, fluency, relevance, and logic)Human evaluators (n=6)
Marmol-Romero et al [20] (2024)Frequency and count of messages per session and gamification participationHuman/Expert (satisfaction, frequency of use, meeting expectations, problems encountered, preferences, and motivations)Not reported
Chen et al [37] (2024)BLEU (BLEU-1, BLEU-2, BLEU-3, BLEU-4), ROUGEe (ROUGE-1, ROUGE-2, ROUGE-L), BERTScoreHuman/Expert (consistency, pertinency, rationality, accuracy, and relevancy)Human evaluation sample size (n=409)
Agnihotri et al [21] (2021)Accuracy and F1-scoreHuman/Expert (emotional relevance and contextual relevance)Not reported
George et al [22] (2024)PsychoBench: Emotional Intelligence Scale and Empathy ScaleNo human evaluationNot reported
Xiao et al [38] (2024)GPT-4 was used to automatically rate the AIf responses in empathy, logical coherence, guidance, and overall scoreHuman/Expert (empathy, coherence, and guidance), User (PANAS [Positive and Negative Affect Schedule] before and after therapy)Human evaluation (n=8)
Kang et al [23] (2025)NLP-based analysis against DSM-5g criteriaHuman/Expert (interviews and open-ended survey), Users Rating (empathy, accuracy, usefulness, complex thinking, emotions, active listening, appropriate questions, positivity, support, professionalism, and personalization)Participants (n=20)
Qiu et al [24] (2024)Accuracy, F1-score, perplexity, METEORh, BLEU-1/2/3, ROUGE-L, and distinct-1/2Human/Expert and user surveys (fluency, coherence, empathy, usefulness, and appropriateness)Not reported
Hanji et al [32]
(2024)
iHuman/Expert (empathy, coherence, sentiment adaptation, and user engagement quality)Not reported
Kang et al [25]
(2024)
ROUGE and fluencyHuman/Expert (readability, professionalism, and match score)Not reported
Fan et al [26]
(2025)
Accuracy, precision, recall, GPT-4 to evaluate (semantic similarity, fluency, expertise, and empathy)Human/Expert: professionalism, fluency, and empathyHuman-related samples (n=200)
Mishra et al [27]
(2023)
Accuracy (W-ACCj) and macro-F1, perplexity, BERTScore-F1, and response lengthHuman/Expert: consistency/correctness (gender-age, persona, psychotherapeutic approach, politeness, interpersonal behavior), fluency, consistency, and nonrepetitivenessHuman experts (n=6) and annotators (n=3)
Gu and Zhu [28] (2024)BLEU score (1, 2, 3, 4, avg), BERTScore (PBERT, RBERT, and FBERT), distinct-2Human/Expert (fluency, helpfulness, relevance, empathy, and professionalism)Human experts (n=6)
Gaikwad et al [33] (2024)Word error rate, accuracy, and perplexityNo human evaluationNot reported
Na et al [36] (2024)Accuracy, recall, F1-score, BLEU, METEOR, CHRFk, BLEURTl, BERTScoreHuman/Expert (Relevance, CBT structure, and helpfulness)Human evaluators (n=6)
Yu and McGuinness et al [29] (2024)Perplexity and BLEUUser/Professional evaluation (perceived utility, usage willingness, human-likeness, supportiveness, and overall rating)Participants (n=20; 10 patients, 10 professionals)
Hu et al [39] (2024)User satisfaction, qualitative feedback on authenticity, entertainment value, fluency, and usefulnessParticipants (n=48) + Interviews (n=10)
Mavila et al [34] (2024)Accuracy, precision, recall, and F1-scoreUser satisfaction (empathy, relevance, and helpfulness)Not reported
Hu et al [30] (2024)Standard, elastic accuracy, ROUGE-1 and ROUGE-L, BLEU-4, BERTScoreHuman/Expert (assess and eliminate low-quality data)Not reported

aNLP: natural language processing.

bBLEU: Bilingual Evaluation Understudy.

cROUGE-L: Recall-Oriented Understudy for Gisting Evaluation Longest Common Subsequence.

dBERTScore: Bidirectional Encoder Representations from Transformers Score.

eROUGE: Recall-Oriented Understudy for Gisting Evaluation.

fAI: artificial intelligence.

gDSM-5: Diagnostic and Statistical Manual of Mental Disorders (Fifth Edition).

hMETEOR: Metric for Evaluation of Translation with Explicit Ordering.

iNot available.

jW-ACC: weighted accuracy.

kCHRF: character n-gram F-score.

lBLEURT: BLEU with Representations from Transformers.

Performance Outcomes

Reported outcomes differed across studies using different architectures, fine-tuning strategies, and evaluation approaches; however, direct performance comparisons were not feasible due to heterogeneity in study design and metrics. Studies that applied task-specific or instruction-based fine-tuning consistently reported stronger outcomes across both human and automated assessments. These models demonstrated higher topic coherence, contextual appropriateness, and emotional responsiveness. For instance, instruction-tuned models were often evaluated using human-rated empathy and relevance criteria, with several studies reporting higher scores relative to their own baselines, zero-shot GPT variants.

Conversely, models evaluated under zero-shot conditions or deployed without domain adaptation tended to struggle with emotionally sensitive content. These models often produced generic or inconsistent responses, with several studies reporting lower scores on fluency, helpfulness, or therapeutic fit. Some models failed to maintain coherent dialogue across multiple turns, especially when faced with ambiguous or distress-laden prompts.

Qualitative findings reinforced these gaps. While several studies noted strengths in fluency, turn-taking, and user engagement, others identified persistent issues such as repetitiveness, surface-level advice, and lack of empathy. User trust and perceived empathy were reported inconsistently; some users described feeling supported, while others perceived the chatbot as impersonal or emotionally disconnected. A few studies highlighted the trade-off between linguistic quality and therapeutic value. Models that generated fluent and natural-sounding text often lacked deeper reflective or validating responses expected in clinical contexts. Only a small subset of studies used hybrid evaluation frameworks that integrated human rubrics (eg, CARE [Consultation and Relational Empathy], supportiveness, and validation) with NLP metrics, suggesting a need for more comprehensive, clinically informed evaluation pipelines.

Performance outcomes across the studies showed notable variation depending on model type, evaluation strategy, and adaptation method. For example, the e-THERAPIST model [27] achieved high automatic evaluation scores, including a gender-age consistency of 90.1%, persona consistency of 84.1%, and psychotherapeutic approach correctness of 92.6%. Its human evaluation scores were similarly high, with fluency rated at 4.62 and consistency at 4.60. The HealMe model [38], designed to identify negative thought patterns, reported a perplexity of 2.52 and a BERTScore (Bidirectional Encoder Representations from Transformers Score)-F1 of 0.89, indicating both lexical and contextual fluency. In comparative benchmarking, PBChat [37] outperformed baseline ChatGLM2 across all major metrics, with ROUGE-L (Recall-Oriented Understudy for Gisting Evaluation Longest Common Subsequence) scores of 28.18 (PanGu) and 23.56 (WenZhong), and distinct-n scores up to 12.74. Helpfulness and fluency in PBChat were rated 3.87 and 4.36, respectively, based on expert review. Meanwhile, models relying solely on zero-shot inference (eg, standard GPT without fine-tuning) tended to underperform, with lower human-rated empathy and coherence, especially in complex or emotionally sensitive scenarios. Despite strong performance in specific metrics, only 3 studies [22,23,38] used psychometrically grounded tools (eg, PHQ-9) to evaluate MH alignment, indicating a persistent gap in clinically validated performance benchmarking.

Service Approach

The reviewed studies explored diverse service models for deploying LLMs. The most common approach was the deployment of standalone chatbot systems, reported in 18 studies [20-37], often via web-based platforms offering real-time conversational support. Prominent examples of standalone deployments include Psy-LLM [35], PBChat [37], TACA [21], Mello [22], HealMe [38], PsyChat [24], MentalBlend [28], Sahara: Virtual Companion [33], CBT-LLM [36], and e-THERAPIST [27]. These systems were typically accessed through browser interfaces or dedicated portals, offering direct user-chatbot interaction without integration into broader ecosystems.

Virtual agents were used in 4 studies (PsyChat [24], Kang et al [25], and PsycoLLM [30], Moon et al [31]), where LLMs were embedded as AI-driven entities within broader service frameworks. These agents were designed to simulate therapist-like roles or guide users through decision support, often involving role-based or persona-enhanced interaction strategies beyond traditional rule-based chatbot systems. Only 2 studies [20,29] implemented embedded chatbot functionality within existing platforms. Yu et al [29] integrated their chatbot into a Unity-based MH simulation (Unity Technologies), while Marmol-Romero et al [20] deployed a Telegram-based chatbot interface (Telegram Messenger LLP) to improve user reach and reduce barriers to entry.

Notably, only one study (Self-Heal [32]) explicitly reported building a mobile app, suggesting limited attention to app-based delivery despite mobile devices being a primary access point for DMHI globally. Importantly, no study used a hybrid deployment strategy (eg, simultaneous deployment via web, mobile, and messaging platforms), highlighting a gap in cross-platform interoperability and multimodal user engagement design. Most of the current implementations remain single-channel, limiting their potential reach and integration into broader digital health ecosystems.

Risk of Bias and Study Quality

Using our structured traffic light quality appraisal across 5 methodological domains (design, dataset reporting, evaluation metrics, external validation, and ethics), most studies demonstrated low risk for dataset reporting and evaluation metrics, suggesting that data sources and performance measures were generally well described (Figures 2 and 3). Design quality was also predominantly rated as low risk, with only a small subset of studies showing some concerns. In contrast, external validation emerged as the most frequent limitation, with several studies rated as some concerns and a few rated as high risk due to limited independent validation or incomplete reporting of validation procedures. Ethics-related reporting showed the greatest variability, with multiple studies rated as high risk and others rated as some concerns, reflecting inconsistent reporting of privacy safeguards, governance oversight, and safety considerations.

Transparency and Reproducibility

Reproducibility remains a significant limitation in LLM-based MH chatbot research. Among the 20 included studies [20-39], only 6 [21,22,24,31,32,38] provided public access to source code or pretrained models, and 4 [21,22,31,38] shared any portion of their datasets. Most studies relied on proprietary or institution-specific data without sufficient documentation to support independent reimplementation. Beyond the limited availability of source code, few studies provided sufficient documentation of preprocessing pipelines, prompt templates, or evaluation protocols to enable full reproducibility. This lack of end-to-end transparency was associated with limited opportunities for independent benchmarking, external validation, and assessment of real-world applicability.

Geographic and Cultural Context

Several studies used region-specific language models and datasets, reflecting diverse linguistic and cultural contexts. For example, ChatGLM and Qwen models were used in Chinese-language applications, while models such as Baichuan and PanGu were also designed for East Asian populations. Despite this, few studies explicitly discussed cultural adaptation strategies or localized evaluation. No study addressed cross-cultural generalizability or model fairness across ethnic or regional groups. Given the sociocultural nuances of MH, studies should emphasize linguistic inclusivity, cultural sensitivity, and population-specific validation in LLM-based applications.

Cross-Study Synthesis

Cross-study synthesis revealed 3 recurring patterns across the included studies. First, evaluation approaches predominantly relied on linguistic or proxy performance metrics, such as automated text similarity scores or rubric-based qualitative ratings, whereas clinically grounded or psychometric outcome measures were infrequently reported. Only a small subset of studies incorporated validated MH instruments, and these were typically applied in research-specific contexts rather than standardized clinical workflows. Second, validation practices were limited across the literature. Few studies conducted external validation using independent datasets or user populations, and reporting of rater calibration, interrater reliability, or structured evaluation protocols was uncommon. Most evaluations were conducted within single-study settings, often using bespoke or internally defined assessment criteria, which constrained comparability across studies. Third, the reporting of safety mechanisms and governance strategies varied substantially. While some studies described the use of prompt constraints, refusal strategies, or rule-based safeguards to mitigate harmful outputs, many provided limited detail on how safety risks were operationally addressed or monitored. Formal descriptions of escalation pathways, bias auditing procedures, or governance frameworks were inconsistently reported across the included studies.

Cost and Resource Reporting

Computational efficiency was rarely described across the included studies. Key deployment details, such as parameter count, hardware specifications, training duration, and inference latency, were often omitted. Only a small subset mentioned resource-saving adaptations, most commonly parameter-efficient fine-tuning. Limited reporting on compute and runtime characteristics restricted the assessment of feasibility and real-world readiness.

Data Availability and Reporting

Not all studies provided complete information for every evaluation criterion. Some studies lacked details on specific aspects such as dataset sources, fine-tuning strategies, tokenization methods, or evaluation metrics. In these cases, the corresponding fields were marked as NA in our analysis. As a result, the total study counts for individual categories did not always sum to 20 (Tables 1 and 2). This reflects variability in reporting standards across studies and highlights the need for more standardized methodologies in LLM-based MH chatbot evaluations.

Emerging Trends and Innovation

Recent trends reveal a shift from static, single-turn chatbots to dynamic multiagent conversational systems and multimodal AI companions. Several studies began incorporating RAG, CoT prompting, and instruction tuning to simulate reasoning and improve empathy. Others explored modular architectures and hybrid agent systems capable of collaborative role-play, such as therapist-patient-coach simulations. Although still in early stages, these advances highlight the growing complexity and potential of next-generation LLM-based DMHIs.


Principal Findings

Our review highlights the increasing adoption of LLM-based models in MH counseling, with a predominant reliance on GPT-based architectures. Despite improvements in natural language understanding and conversational adaptability, several limitations persist in current implementations. The lack of clinical benchmarking raises concerns about the real-world applicability and safety of these models. The overwhelming use of proprietary datasets restricts reproducibility, making it difficult to generalize findings across diverse populations. Furthermore, evaluation frameworks remain inconsistent, with studies relying on general NLP performance metrics rather than clinically meaningful outcome measures. While fine-tuning strategies were used in several studies, their impact on response accuracy and clinical relevance remains underexplored. Fine-tuned models demonstrated better alignment with MH-specific dialogues, but the trade-offs in computational cost and dataset limitations restricted their scalability. The effectiveness of prompting techniques also varied, with few-shot prompting outperforming zero-shot prompting in maintaining contextual awareness and improving response coherence. However, CoT prompting, which could enhance multistep reasoning in counseling interactions, was rarely used.

Study Limitations

Several methodological limitations should be acknowledged. First, the search was limited to English-language publications, which may have excluded relevant non-English work. Second, the relatively small number of eligible studies and their methodological heterogeneity constrained the generalizability of our findings. Third, due to variations in modeling approaches and outcome reporting, a formal meta-analysis could not be performed, and the synthesis remains descriptive rather than quantitative. Finally, interstudy variability in reporting model architecture and evaluation metrics constrained comparability across results. Future systematic reviews could address these issues by expanding database coverage, including multilingual evidence, and conducting meta-analytic modeling once more homogeneous data become available.

Regulatory Readiness and Clinical Integration

From a counseling practice perspective, the predominance of technical performance metrics over clinically grounded outcomes limits the interpretability of reported effectiveness for real-world MH care. Without standardized psychometric evaluation, external validation, or integration into established care pathways, current LLM-based counseling systems remain difficult to assess in terms of therapeutic appropriateness, risk management, and clinical utility. These limitations may hinder clinician trust and slow responsible integration into digital MH service delivery.

None of the reviewed studies reported formal clinical validation, trial registration, or engagement with regulatory bodies such as the Food and Drug Administration (FDA). While a few referenced potential deployment in counseling or therapy contexts, no study implemented the chatbot in a controlled clinical workflow or evaluated clinical efficacy using standard instruments. Integration into digital MH care systems therefore remains conceptual. Bridging this gap will require interdisciplinary collaboration with clinicians, formal usability testing, and alignment with medical device regulations and health data standards.

Although ethical risks such as bias, hallucinated outputs, and inappropriate reassurance were frequently acknowledged, few studies provided operational detail on how these risks were monitored or mitigated. Formal mechanisms for handling high-risk disclosures, escalation to human professionals, or postdeployment monitoring were rarely described. Moreover, explicit engagement with regulatory pathways, clinical oversight models, or medical device governance frameworks was largely absent. These gaps raise concerns regarding regulatory readiness and highlight the need for stronger ethical and clinical infrastructure prior to large-scale deployment in MH settings.

Cost and Resource Efficiency

Most studies did not report details on model size, inference latency, or computational requirements. Only a few mentioned using parameter-efficient fine-tuning techniques such as LoRA to reduce training cost. This absence of cost-awareness limits the practical deployability of many proposed models, particularly in resource-constrained operational environments. In clinical settings, hospitals and MH clinics, infrastructure budgets are often tightly constrained and must balance multiple priorities. High resource demands may hinder scalability or lead to reliance on external cloud services, which raises further concerns about data privacy, latency, and long-term operational cost. Future work should explicitly report hardware constraints, training time, model size, and inference cost for potential use in low- and middle-income countries or high-demand clinical settings where efficient and secure solutions are essential.

Comparison With Prior Work

While prior systematic reviews have explored various aspects of LLM-driven MH applications, crucial gaps remain in assessing their real-world utility and clinical effectiveness. Studies such as Guo et al [10] and Hua et al [9] examined LLM applicability in MH but provided limited assessments of conversational interventions. Other reviews focused on safety [40] and ethical concerns [41,42], yet lacked discussions on practical deployment, user adoption, and chatbot-specific challenges remain scarce. Additionally, existing studies lack standardized evaluation frameworks for assessing MH chatbot interactions, with few exceptions, such as Abbasian et al [43], proposing structured metrics that remain largely unvalidated in clinical settings. Despite the increasing adoption of LLMs in DMHI, there is no systematic comparison of fine-tuning techniques, prompt engineering, or model-specific adaptations for MH chatbots.

Compared to earlier literature that often treated LLMs as monolithic tools or black-box models, recent studies [44,45] have begun refining their roles with DMHI through structured prompt design, domain adaptations, explainable LLM, and human-in-the-loop feedback. However, technological advances continue to emerge rapidly, while a gap between experimental prototypes and deployable, interpretable, patient-safe tools persists. This systematic review contributes to the field by systematically analyzing 20 LLM-based MH chatbots across deployment models, prompting strategies, and evaluation designs, assessing their real-world adoption and accessibility.

Limitations and Directions for Future Modeling

A key limitation across current LLM-based MH chatbots is limited trustworthiness and interpretability. Many systems rely on black-box generative models without incorporating explicit explainability mechanisms, raising concerns regarding user confidence and the safety of automated MH support. Although most models enable real-time conversational interaction, their adaptive capabilities remain constrained. Rather than dynamically responding to evolving user context, most systems operate through predefined conversational structures. Personalization is similarly limited, as few models incorporate user history, sentiment signals, or behavioral context to inform tailored responses. A small number of studies suggest the potential value of more adaptive or multimodal designs; however, existing evidence remains preliminary and insufficient to establish their effectiveness or clinical viability.

Despite growing interest in evaluation, substantial gaps in performance reporting continue to limit cross-study comparison and generalizability. While several studies referenced user satisfaction, most reported it as a binary outcome without quantifying engagement depth or therapeutic relevance. Although the SUS was mentioned in a small number of studies, none reported numeric scores, constraining benchmarking of usability and acceptability. Expert evaluation frameworks were inconsistently described, with limited reporting on rating dimensions, scale definitions, or interrater reliability. Response time, a critical usability factor, was occasionally noted but not systematically measured. Some studies introduced internally defined scoring schemes, such as the Emotional Impact Score (EIS), but these lacked standardized definitions. Collectively, these inconsistencies underscore the need for more structured, transparent, and clinically grounded evaluation pipelines.

In addition, most studies did not disclose detailed information regarding training data provenance or fine-tuning procedures, limiting transparency and reproducibility. While several works explored increasingly complex system designs, formal safeguards related to ethical governance, secure data handling, and clinical oversight were rarely specified. Advancing toward clinically deployable LLM-based MH systems will require a rigorous external validation framework, standardized outcome measurement, and explicit attention to regulatory and ethical alignment.

Taken together, current evidence indicates that LLM-based counseling systems remain at an early developmental stage characterized by rapid technical experimentation but limited empirical grounding. Progress in this domain will depend on the adoption of standardized evaluation frameworks that integrate both NLP-based and psychometric measures, transparent reporting of model development and validation practices, and multidisciplinary collaboration among clinicians, data scientists, and ethicists to support safe, equitable, and clinically accountable deployment.

Conclusions

This systematic review highlights the growing use of LLM-based chatbots in MH counseling. While these models show promise in enhancing accessibility and response adaptability, significant challenges remain, including a lack of clinical validation, ethical safeguards, and standardized evaluation frameworks. Addressing these gaps will be critical in transitioning LLM-based chatbots from experimental tools to clinically trusted interventions.

Acknowledgments

Generative artificial intelligence tools were used only for language editing; all screening, extraction, and interpretation were performed by the authors.

Funding

No external financial support or grants were received from any public, commercial, or not-for-profit entities for the research, authorship, or publication of this article.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Search queries.

DOCX File , 14 KB

Multimedia Appendix 2

Glossary of key terms.

DOCX File , 13 KB

Multimedia Appendix 3

PRISMA checklist.

PDF File (Adobe PDF File), 1875 KB

  1. Pryor TA, Gardner RM, Clayton PD, Warner HR. The HELP system. J Med Syst. 1983;7(2):87-102. [CrossRef] [Medline]
  2. Gardner RM, Golubjatnikov OK, Laub RM, Jacobson JT, Evans RS. Computer-critiqued blood ordering using the HELP system. Comput Biomed Res. 1990;23(6):514-528. [CrossRef] [Medline]
  3. Wainberg ML, Scorza P, Shultz JM, Helpman L, Mootz JJ, Johnson KA, et al. Challenges and opportunities in global mental health: a research-to-practice perspective. Curr Psychiatry Rep. 2017;19(5):28. [FREE Full text] [CrossRef] [Medline]
  4. Werntz A, Amado S, Jasman M, Ervin A, Rhodes JE. Providing human support for the use of digital mental health interventions: systematic meta-review. J Med Internet Res. 2023;25:e42864. [FREE Full text] [CrossRef] [Medline]
  5. Vaidyam AN, Wisniewski H, Halamka JD, Kashavan MS, Torous JB. Chatbots and conversational agents in mental health: a review of the psychiatric landscape. Can J Psychiatry. 2019;64(7):456-464. [FREE Full text] [CrossRef] [Medline]
  6. Abd-Alrazaq AA, Alajlani M, Alalwan AA, Bewick BM, Gardner P, Househ M. An overview of the features of chatbots in mental health: a scoping review. Int J Med Inform. 2019;132:103978. [FREE Full text] [CrossRef] [Medline]
  7. Kolding S, Lundin RM, Hansen L, Østergaard SD. Use of generative artificial intelligence (AI) in psychiatry and mental health care: a systematic review. Acta Neuropsychiatr. 2024;37:e37. [CrossRef] [Medline]
  8. Yuan A, Garcia Colato E, Pescosolido B, Song H, Samtani S. Improving workplace well-being in modern organizations: a review of large language model-based mental health chatbots. ACM Trans Manage Inf Syst. 2025;16(1):1-26. [FREE Full text] [CrossRef]
  9. Hua Y, Liu F, Yang K, Li Z, Na H, Sheu YH, et al. Large language models in mental health care: a scoping review. Curr Treat Options Psych. 2025;12(1). [CrossRef]
  10. Guo Z, Lai A, Thygesen JH, Farrington J, Keen T, Li K. Large language models for mental health applications: systematic review. JMIR Ment Health. 2024;11:e57400. [FREE Full text] [CrossRef] [Medline]
  11. Triantafyllopoulos A, Terhorst Y, Tsangko I, et al. Large language models for mental health. arXiv. Preprint posted online on Nov 4 2024. [FREE Full text]
  12. Lawrence HR, Schneider RA, Rubin SB, Matarić MJ, McDuff DJ, Jones Bell M. The opportunities and risks of large language models in mental health. JMIR Ment Health. 2024;11:e59479. [FREE Full text] [CrossRef] [Medline]
  13. Han T, Adams L, Papaioannou JM, Grundmann P, Oberhauser T, Figueroa A, et al. MedAlpaca: an open-source collection of medical conversational AI models and training data. arXiv. Preprint posted online on April 14 2023. [FREE Full text]
  14. Chen L, Preece DA, Sikka P, Gross J, Krause B. A framework for evaluating appropriateness, trustworthiness, and safety in mental wellness AI chatbots. arXiv. Preprint posted online on July 16 2024. [FREE Full text] [Medline]
  15. Yang K, Zhang T, Kuang Z, Xie Q, Huang J, Ananiadou S. MentaLLaMA: interpretable mental health analysis on social media with large language models. ACM; 2024. Presented at: WWW '24: Proceedings of the ACM Web Conference 2024; 2024 May, 13-17:4489-4500; Singapore. [CrossRef]
  16. Jiang AQ, Sablayrolles A, Mensch A. Mistral 7B. arXiv. Preprint posted online on October 10 2023. [FREE Full text]
  17. Shaoxiong J, Tianlin Z, Luna A, Jie F, Prayag T, Erik C. MentalBERT: publicly available pretrained language models for mental healthcare. European Language Resources Association; 2022. Presented at: Proceedings of the Thirteenth Language Resources and Evaluation Conference; January 28, 2026:7184-7190; Marseille, France. URL: https://aclanthology.org/2022.lrec-1.778/
  18. Ke L, Tong S, Cheng P, Peng K. Exploring the frontiers of LLMs in psychological applications: a comprehensive review. Artif Intell Rev. 2025;58(10). [CrossRef]
  19. Li Y, Li Z, Zhang K, Dan R, Jiang S, Zhang Y. ChatDoctor: a medical chat model fine-tuned on a large language model meta-AI (LLaMA) using medical domain knowledge. Cureus. 2023;15(6):e40895. [FREE Full text] [CrossRef] [Medline]
  20. Mármol-Romero AM, García-Vega M, García-Cumbreras MAA, Montejo-Ráez A. An empathic GPT-based chatbot to talk about mental disorders with Spanish teenagers. Int J Hum Comput Interact. 2024;40:1-17. [CrossRef]
  21. Agnihotri M, Pooja Rao SB, Jayagopi DB, Hebbar S, Rasipuram S, Maitra A, et al. Towards generating topic-driven and affective responses to assist mental wellness. Springer; 2021. Presented at: Pattern Recognition. ICPR International Workshops and Challenges; January 10-15, 2021:129-143; Virtual Event. [CrossRef]
  22. George SB, Binu RM, Ebin P. Mello: a large language model for mental health counselling conversations. IEEE; 2024. Presented at: 3rd International Conference for Advancement in Technology (ICONAT); September 6-8, 2024:1-6; Goa, India. [CrossRef]
  23. Kang B, Hong M. Development and evaluation of a mental health chatbot using chatgpt 4.0: mixed methods user experience study with Korean users. JMIR Med Inform. 2025;13:e63538. [FREE Full text] [CrossRef] [Medline]
  24. Qiu H, Li A, Ma L, Lan Z. PsyChat: a client-centric dialogue system for mental health support. Association for Computing Machinery; 2024. Presented at: 27th International Conference on Computer Supported Cooperative Work in Design (CSCWD); May 8-10, 2024:2979-2984; Tianjin, China. [CrossRef]
  25. Kang C, Novak D, Urbanova K, Cheng Y, Hu Y. Domain-specific improvement on psychotherapy chatbot using assistant. IEEE; 2024. Presented at: IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW); April 14-19, 2024; Seoul, Republic of Korea. [CrossRef]
  26. Fan X, Yang L, Wang X, Lyu D, Chen H. Constructing a knowledge-guided mental health chatbot with LLMs. 2025. Presented at: Proceedings of the 16th Asian Conference on Machine Learning; December 5-8, 2024:287-302; Hanoi, Vietnam. [CrossRef]
  27. Mishra K, Priya P, Burja M, Ekbal A. e-THERAPIST: I suggest you cultivate a mindset of positivity and nurture uplifting thoughts. Association for Computational Linguistics; 2023. Presented at: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing; January 28, 2026:13952-13967; Singapore. URL: https://aclanthology.org/2023.emnlp-main.861/ [CrossRef]
  28. Gu Z, Zhu Q. MentalBlend: enhancing online mental health support through the integration of LLMs with psychological counseling theories. 2024. Presented at: Proceedings of the Annual Meeting of the Cognitive Science Society; July 24–27, 2024; Rotterdam, The Netherlands. URL: https://escholarship.org/uc/item/7dk883nx
  29. Yu HQ, McGuinness S. An experimental study of integrating fine-tuned large language models and prompts for enhancing mental health support chatbot system. J Med Artif Intell. 2024;7:16. [CrossRef]
  30. Hu J, Dong T, Luo G, Ma H, Zou P, Sun X, et al. PsycoLLM: Enhancing LLM for psychological understanding and evaluation. IEEE Trans Comput Soc Syst. 2025;12(2):539-551. [CrossRef]
  31. Moon P, Bhattacharyya P. We Care: multimodal depression detection and knowledge-infused mental health therapeutic response generation. NLP Association of India (NLPAI); 2024. Presented at: Proceedings of the 21st International Conference on Natural Language Processing (ICON); January 28, 2026:296-310; AU-KBC Research Centre, Chennai, India. URL: https://aclanthology.org/2024.icon-1.35/
  32. Hanji BR, Charanya S, Gupta D, Krishna E, Devadiga HG. Self-Heal: conversational therapy bot with AI-enhanced features for mental health. IEEE; 2024. Presented at: 8th International Conference on Computational System and Information Technology for Sustainable Solutions (CSITSS); November 7-9, 2024:1-6; Bengaluru, India. [CrossRef]
  33. Gaikwad S, Iyer AB, Talluri KS, Salve P. Sahara: Virtual companion - exploring multimodal empathetic conversational AI leveraging ensemble learning with humor. IJSR. 2024;13(4):1501-1507. [CrossRef]
  34. Mavila R, Jaiswal S, Naswa R, Yuwen W, Erdly B, Si D. iCare: an AI-powered virtual assistant for mental health. IEEE; 2024. Presented at: IEEE 12th International Conference on Healthcare Informatics (ICHI); June 3-6, 2024:466-471; Orlando, FL, USA. [CrossRef]
  35. Lai T, Shi Y, Du Z, Wu J, Fu K, Dou Y, et al. Supporting the demand on mental health services with AI-based conversational large language models (LLMs). BioMedInformatics. 2023;4(1):8-33. [CrossRef]
  36. Na H. CBT-LLM: a Chinese large language model for cognitive behavioral therapy-based mental health question answering. ELRA and ICCL; 2024. Presented at: Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024); January 28, 2026:2930-2940; Torino, Italia.
  37. Chen P, Fan Z, Lu Y, Xu Q. PBChat: enhance student's problem behavior diagnosis with large language model. Springer Nature; 2024. Presented at: Artificial Intelligence in Education 25th International Conference, AIED 2024; July 8-12, 2024:32-45; Recife, Brazil. [CrossRef]
  38. Xiao M, Xie Q, Kuang Z, Liu Z, Yang K, Peng M, et al. HealMe: harnessing cognitive reframing in large language models for psychotherapy. Association for Computational Linguistics; 2024. Presented at: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); August 11-16, 2024:1707-1725; Bangkok, Thailand. [CrossRef]
  39. Hu Z, Hou H, Ni S. Grow with your AI buddy: designing an LLMs-based conversational agent for the measurement and cultivation of children's mental resilience. Association for Computing Machinery; 2024. Presented at: IDC '24: Proceedings of the 23rd Annual ACM Interaction Design and Children Conference; June 17-20, 2024:811-817; Delft Netherlands. [CrossRef]
  40. De Freitas J, Uğuralp AK, Oğuz‐Uğuralp Z, Puntoni S. Chatbots and mental health: insights into the safety of generative AI. J Consum Psychol. 2023;34(3):481-491. [CrossRef]
  41. Cabrera J, Loyola MS, Magana I, Rojas R. Ethical dilemmas, mental health, artificial intelligence, and LLM-based chatbots. Springer Nature; 2023. Presented at: Bioinformatics and Biomedical Engineering 10th International Work-Conference, IWBBIO 2023; July 12-14, 2023:313-326; Meloneras, Gran Canaria, Spain. [CrossRef]
  42. Rahsepar Meadi M, Sillekens T, Metselaar S, van Balkom A, Bernstein J, Batelaan N. Exploring the ethical challenges of conversational AI in mental health care: scoping review. JMIR Ment Health. 2025;12:e60432. [FREE Full text] [CrossRef] [Medline]
  43. Abbasian M, Khatibi E, Azimi I, Oniani D, Shakeri Hossein Abad Z, Thieme A, et al. Foundation metrics for evaluating effectiveness of healthcare conversations powered by generative AI. NPJ Digit Med. 2024;7(1):82. [FREE Full text] [CrossRef] [Medline]
  44. Elsweiler D, Meyer S. Evaluating the efficacy, controllability, and safety of LLM-driven conversational agents to support behaviour change. arXiv. Preprint posted online on August 7 2023. [FREE Full text]
  45. Ma, Xiao, Mishra, Shubham, Liu, Annie, Su, S. Y, Chen, Jian, Kulkarni, Chinmay, et al. Cheng, H. T.. Beyond chatbots: ExploreLLM for structured thoughts and personalized model responses. Association for Computing Machinery; 2024. Presented at: CHCHI EA ’24: Extended Abstracts of the CHI Conference on Human Honolulu, HI, USA in Computing SystemsI EA '24: Extended Abstracts of the CHI Conference on Human Factors in Computing Systems; May 11–16, 2024; Honolulu HI USA.


AI: artificial intelligence
BERT: Bidirectional Encoder Representations from Transformers
BERTScore: Bidirectional Encoder Representations from Transformers Score
BLEU: Bilingual Evaluation Understudy
CARE: Consultation and Relational Empathy
CBT: cognitive behavioral therapy
CoT: chain-of-thought
DMHI: digital mental health intervention
EIS: Emotional Impact Score
FDA: Food and Drug Administration
GAD-7: Generalized Anxiety Disorder-7
LLM: large language model
LoRA: low-rank adaptation
MH: mental health
NLP: natural language processing
OECD: Organisation for Economic Cooperation and Development
PEFT: parameter-efficient fine-tuning
PHQ-9: Patient Health Questionnaire-9
PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses
QLoRA: quantized low-rank adaptation
RAG: retrieval-augmented generation
RLHF: reinforcement learning with human feedback
ROUGE: Recall-Oriented Understudy for Gisting Evaluation
ROUGE-L: Recall-Oriented Understudy for Gisting Evaluation Longest Common Subsequence
SUS: System Usability Scale
WHO: World Health Organization


Edited by A Coristine; submitted 09.Jul.2025; peer-reviewed by R Kornfield, D Reichenpfader; comments to author 01.Sep.2025; accepted 27.Jan.2026; published 13.Mar.2026.

Copyright

©Ha Na Cho, Jiayuan Wang, Di Hu, Kai Zheng. Originally published in JMIR AI (https://ai.jmir.org), 13.Mar.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR AI, is properly cited. The complete bibliographic information, a link to the original publication on https://www.ai.jmir.org/, as well as this copyright and license information must be included.